16 research outputs found

    Incorporating feature ranking and evolutionary methods for the classification of high-dimensional DNA microarray gene expression data

    Get PDF
    Background: DNA microarray gene expression classification poses a challenging task to the machine learning domain. Typically, the dimensionality of gene expression data sets could go from several thousands to over 10,000 genes. A potential solution to this issue is using feature selection to reduce the dimensionality. Aim The aim of this paper is to investigate how we can use feature quality information to improve the precision of microarray gene expression classification tasks. Method: We propose two evolutionary machine learning models based on the eXtended Classifier System (XCS) and a typical feature selection methodology. The first one, which we call FS-XCS, uses feature selection for feature reduction purposes. The second model is GRD-XCS, which uses feature ranking to bias the rule discovery process of XCS. Results: The results indicate that the use of feature selection/ranking methods is essential for tackling high-dimensional classification tasks, such as microarray gene expression classification. However, the results also suggest that using feature ranking to bias the rule discovery process performs significantly better than using the feature reduction method. In other words, using feature quality information to develop a smarter learning procedure is more efficient than reducing the feature set. Conclusion: Our findings have shown that extracting feature quality information can assist the learning process and improve classification accuracy. On the other hand, relying exclusively on the feature quality information might potentially decrease the classification performance (e.g., using feature reduction). Therefore, we recommend a hybrid approach that uses feature quality information to direct the learning process by highlighting the more informative features, but at the same time not restricting the learning process to explore other features

    Incorporating feature ranking and evolutionary methods for the classification of high-dimensional DNA microarray gene expression data

    Get PDF
    BackgroundDNA microarray gene expression classification poses a challenging task to the machine learning domain. Typically, the dimensionality of gene expression data sets could go from several thousands to over 10,000 genes. A potential solution to this issue is using feature selection to reduce the dimensionality.AimThe aim of this paper is to investigate how we can use feature quality information to improve the precision of microarray gene expression classification tasks. Method  We propose two evolutionary machine learning models based on the eXtended Classifier System (XCS) and a typical feature selection methodology. The first one, which we call FS-XCS, uses feature selection for feature reduction purposes. The second model is GRD-XCS, which uses feature ranking to bias the rule discovery process of XCS.ResultsThe  results   indicate   that  the  use  of   feature  selection / ranking methods is essential for tackling high-dimensional classification tasks, such as microarray gene expression classification. However, the results also suggest that using feature ranking to bias the rule discovery process performs significantly better than using the feature reduction method. In other words, using feature quality information to develop a smarter learning procedure is more efficient than reducing the feature set. ConclusionOur findings have shown that extracting feature quality information can assist the learning process and improve classification accuracy. On the other hand, relying exclusively on the feature quality information might potentially decrease the classification performance (e.g., using feature reduction). Therefore, we recommend a hybrid approach that uses feature quality information to direct the learning process by highlighting the more informative features, but at the same time not restricting the learning process to explore other features

    Experimental and theoretical justifications for the observed discriminations between enantiomers of prochiral alcohols by chirally blind EI-MS

    Full text link
    To all appearances, electron impact mass spectrometer (EI-MS) is considered a "chirally blind" instrument. Yet, numerous non-identical R (right) and S (left) configurations of prochiral alcohols' mass spectra alcohols have appeared in the literature with almost no justification. Such observations are often attributed to impurities, experimental circumstances, inaccurate measurements, etc. In an experimental attempt to explain this phenomenon, here we have avoided the above mentioned pitfalls by conducting control experiments using different pure enantiomers under the same circumstances. Hence, we report the mass spectra of R- and S-enantiomers of 2-octanol (1R, 1S) and 1-octyn-3-ol (2R, 2S) collected by running 20 independent experiments for each R- and S-enantiomer. Statistical analyses confirmed that the peak intensities were significant to an acceptable level of confidence. The 1R and 1S enantiomers were separated reasonably in the PC space, implying that the chirally blind EI-MS is able to discriminate between R and S prochiral alcohols. Theoretically, self-complexation through H-bonding for S (or R) appears to give a new chiral center at the H-bonded oxygen atom, producing a new dimeric pair of diastereomers SRS and SSS (or RRR and RSR) before ionization, and SRS.+ and SSS.+ (or RRR.+ and RSR.+) after ionization. The results of our calculations have explicitly shown that these hydrogen bonds formed. Interestingly, the latter four ionized diastereomers appear with different structural and thermodynamic parameters at the M06-2X/6-311++g (d,p) level of theory

    High performance computing enabling exhaustive analysis of higher order single nucleotide polymorphism interaction in Genome Wide Association Studies.

    Get PDF
    Genome-wide association studies (GWAS) are a common approach for systematic discovery of single nucleotide polymorphisms (SNPs) which are associated with a given disease. Univariate analysis approaches commonly employed may miss important SNP associations that only appear through multivariate analysis in complex diseases. However, multivariate SNP analysis is currently limited by its inherent computational complexity. In this work, we present a computational framework that harnesses supercomputers. Based on our results, we estimate a three-way interaction analysis on 1.1 million SNP GWAS data requiring over 5.8 years on the full "Avoca" IBM Blue Gene/Q installation at the Victorian Life Sciences Computation Initiative. This is hundreds of times faster than estimates for other CPU based methods and four times faster than runtimes estimated for GPU methods, indicating how the improvement in the level of hardware applied to interaction analysis may alter the types of analysis that can be performed. Furthermore, the same analysis would take under 3 months on the currently largest IBM Blue Gene/Q supercomputer "Sequoia" at the Lawrence Livermore National Laboratory assuming linear scaling is maintained as our results suggest. Given that the implementation used in this study can be further optimised, this runtime means it is becoming feasible to carry out exhaustive analysis of higher order interaction studies on large modern GWAS.This research was partially funded by NHMRC grant 1033452 and was supported by a Victorian Life Sciences Computation Initiative (VLSCI) grant number 0126 on its Peak Computing Facility at the University of Melbourne, an initiative of the Victorian Government, Australia

    GPU-accelerated eXtended classifier system

    No full text
    XCS - the extended Classifier System - combines an evolutionary algorithm with reinforcement learning to evolve a population of condition-action rules (classifiers). Typically, population-based approaches are slow and increasing the problem size (in terms of the number of features/samples) poses a real threat to the suitability of XCS for real-world applications. Thus, reducing the execution time without losing accuracy is highly desirable. Profiling of the execution of off-the-shelf XCS implementations suggests that the rule matching process is the most computational demanding step. A solution to this is parallelization, i.e., using parallel processing techniques to speed up the matching process (and thus the entire XCS learning process). There are many ways to achieve that, using Graphic Processing Units (GPUs) is one option. Originally, GPUs were designed to conduct a sequence of graphics operations in a massively parallel fashion. Today, GPUs can be used for all sorts of general purpose calculations that are normally handled by the CPU. In this paper, we propose a hybrid rule matching process using both CPU and GPU simultaneously for a maximum performance gain. Our experimental results indicate that this approach does speed up the XCS learning process, and that the GPU is the dominant powerful computing resource in the model
    corecore